SGEM: stochastic gradient with energy and momentum
نویسندگان
چکیده
In this paper, we propose SGEM, stochastic gradient with energy and momentum, to solve a class of general non-convex optimization problems, based on the AEGD method introduced in (adaptive descent energy) Liu Tian (Numerical Algebra, Control Optimization, 2023). SGEM incorporates both momentum so as inherit their dual advantages. We show that features an unconditional stability property provide positive lower threshold for variable. further derive energy-dependent convergence rates setting, well regret bound online convex setting. Our experimental results converges faster than generalizes better or at least SGDM training some deep neural networks.
منابع مشابه
Momentum and Stochastic Momentum for Stochastic Gradient, Newton, Proximal Point and Subspace Descent Methods
In this paper we study several classes of stochastic optimization algorithms enriched with heavy ball momentum. Among the methods studied are: stochastic gradient descent, stochastic Newton, stochastic proximal point and stochastic dual subspace ascent. This is the first time momentum variants of several of these methods are studied. We choose to perform our analysis in a setting in which all o...
متن کاملFast Stochastic Variance Reduced Gradient Method with Momentum Acceleration for Machine Learning
Recently, research on accelerated stochastic gradient descentmethods (e.g., SVRG) has made exciting progress (e.g., lin-ear convergence for strongly convex problems). However,the best-known methods (e.g., Katyusha) requires at leasttwo auxiliary variables and two momentum parameters. Inthis paper, we propose a fast stochastic variance reductiongradient (FSVRG) method...
متن کاملStochastic gradient method with accelerated stochastic dynamics
In this paper, we propose a novel technique to implement stochastic gradient methods, which are beneficial for learning from large datasets, through accelerated stochastic dynamics. A stochastic gradient method is based on mini-batch learning for reducing the computational cost when the amount of data is large. The stochasticity of the gradient can be mitigated by the injection of Gaussian nois...
متن کاملUnsupervised learning with stochastic gradient
A stochastic gradient is formulated based on deterministic gradient augmented with Cauchy simulated annealing capable to reach a global minimum with a convergence speed significantly faster when simulated annealing is used alone. In order to solve space-time variant inverse problems known as blind source separation, a novel Helmholtz free energy contrast function, H 1⁄4 E T0S; with imposed ther...
متن کاملStochastic Gradient Descent with GPGPU
We show how to optimize a Support Vector Machine and a predictor for Collaborative Filtering with Stochastic Gradient Descent on the GPU, achieving 1.66 to 6-times accelerations compared to a CPUbased implementation. The reference implementations are the Support Vector Machine by Bottou and the BRISMF predictor from the Netflix Prices winning team. Our main idea is to create a hash function of ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Numerical Algorithms
سال: 2023
ISSN: ['1017-1398', '1572-9265']
DOI: https://doi.org/10.1007/s11075-023-01621-x